Goto

Collaborating Authors

 evaluation procedure





Improving LLM's Attachment to External Knowledge In Dialogue Generation Tasks Through Entity Anonymization

Sheikhi, Hadi, Huang, Chenyang, Zaïane, Osmar R.

arXiv.org Artificial Intelligence

Knowledge graph-based dialogue generation (KG-DG) is a challenging task requiring models to effectively incorporate external knowledge into conversational responses. While large language models (LLMs) have achieved impressive results across various NLP tasks, their ability to utilize external knowledge in KG-DG remains under-explored. We observe that LLMs often rely on internal knowledge, leading to detachment from provided knowledge graphs, even when they are given a flawlessly retrieved knowledge graph. First, we introduce LLM-KAT, an evaluation procedure for measuring knowledge attachment in generated responses. Second, we propose a simple yet effective entity anonymization technique to encourage LLMs to better leverage external knowledge. Experiments on the OpenDialKG dataset demonstrate that our approach improves LLMs' attachment on external knowledge.




replace the

Neural Information Processing Systems

We thank the reviewers for their valuable input on how to improve our manuscript. We use our evaluation procedure ( 4) since we will not have ground-truth outcomes. The revision will provide this discussion with relevant citations. We would like to clarify that Theorem 3.1 describes the conditions under which our method is optimal. The RF estimation error dominates the confounding error.


ClonEval: An Open Voice Cloning Benchmark

Christop, Iwona, Kuczyński, Tomasz, Kubis, Marek

arXiv.org Artificial Intelligence

ABSTRACT We present a new benchmark for voice cloning text-to-speech models. The benchmark consists of an evaluation protocol, an open-source library for assessing the performance of voice cloning models, and an accompanying leaderboard. The paper discusses design considerations and presents a detailed description of the evaluation procedure. The usage of the software library is explained, along with the organization of the leaderboard. The evaluation results of selected open-source models are reported.


Beyond algorithm hyperparameters: on preprocessing hyperparameters and associated pitfalls in machine learning applications

Sauer, Christina, Boulesteix, Anne-Laure, Hanßum, Luzia, Hodiamont, Farina, Bausewein, Claudia, Ullmann, Theresa

arXiv.org Machine Learning

Adequately generating and evaluating prediction models based on supervised machine learning (ML) is often challenging, especially for less experienced users in applied research areas. Special attention is required in settings where the model generation process involves hyperparameter tuning, i.e. data-driven optimization of different types of hyperparameters to improve the predictive performance of the resulting model. Discussions about tuning typically focus on the hyperparameters of the ML algorithm (e.g., the minimum number of observations in each terminal node for a tree-based algorithm). In this context, it is often neglected that hyperparameters also exist for the preprocessing steps that are applied to the data before it is provided to the algorithm (e.g., how to handle missing feature values in the data). As a consequence, users experimenting with different preprocessing options to improve model performance may be unaware that this constitutes a form of hyperparameter tuning - albeit informal and unsystematic - and thus may fail to report or account for this optimization. To illuminate this issue, this paper reviews and empirically illustrates different procedures for generating and evaluating prediction models, explicitly addressing the different ways algorithm and preprocessing hyperparameters are typically handled by applied ML users. By highlighting potential pitfalls, especially those that may lead to exaggerated performance claims, this review aims to further improve the quality of predictive modeling in ML applications.


A Dataset for Evaluating LLM-based Evaluation Functions for Research Question Extraction Task

Fujisaki, Yuya, Takagi, Shiro, Asoh, Hideki, Kumagai, Wataru

arXiv.org Artificial Intelligence

The progress in text summarization techniques has been remarkable. However the task of accurately extracting and summarizing necessary information from highly specialized documents such as research papers has not been sufficiently investigated. We are focusing on the task of extracting research questions (RQ) from research papers and construct a new dataset consisting of machine learning papers, RQ extracted from these papers by GPT-4, and human evaluations of the extracted RQ from multiple perspectives. Using this dataset, we systematically compared recently proposed LLM-based evaluation functions for summarizations, and found that none of the functions showed sufficiently high correlations with human evaluations. We expect our dataset provides a foundation for further research on developing better evaluation functions tailored to the RQ extraction task, and contribute to enhance the performance of the task. The dataset is available at https://github.com/auto-res/PaperRQ-HumanAnno-Dataset.